Privacy Preserving Machine Learning: Related Work

نویسنده

  • Bargav Jayaraman
چکیده

A practical scenario of PPML is where only one central party has the entire data on which the ML algorithm has to be learned. Agrawal and Ramakrishnan [1] proposed the first method to learn a Decision Tree classifier on a database without revealing any information about individual records. They consider public model private data setting where the algorithm and its parameters are public whereas the data, over which the algorithm is trained, is kept private. The privacy of individual records in data is achieved through data perturbation. This notion of privacy was later identified as differential privacy by Dwork [2]. Though the data is private in the scenario considered by Agrawal and Ramakrishnan, where algorithm is trained on the data and only the learned algorithm is revealed, some orthogonal scenarios consider publishing the perturbed data thereby leading to public data setting with respect to the perturbed data. Chaudhuri and Monteleoni [3] proposed learning of Logistic Regression with differential privacy in which they consider the same setting of public model private data as above. However, a major distinction is that they perturb the learning algorithm rather than the data itself (unlike the method of Agrawal and Ramakrishnan) thereby not revealing information about individual data points upon which the algorithm is learned. Chaudhuri et al. [4] further generalized this solution for learning Support Vector Machine (SVM) under the same settings. Zhang et al. [5] gave a differential privacy solution to learning Logistic and Linear Regression models by perturbing the objective function similar to Chaudhuri and Monteleoni under the public model private data setting. Rubinstein et al. [6] came up with differentially private learning of SVM under public model private data setting, where differential privacy is guaranteed by introducing weight regularization in objective loss function. Jain et al. [7] proposed a mechanism of training Deep Belief Networks with differential privacy using drop-out (unlike the above methods of objective perturbation which rely on regularization for differential privacy). Like all the above differential privacy mechanisms, Jain et al. also consider public model private data setting.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Private Collaborative Neural Network Learning

Machine learning algorithms, such as neural networks, create better predictive models when having access to larger datasets. In many domains, such as medicine and finance, each institute has only access to limited amounts of data, and creating larger datasets typically requires collaboration. However, there are privacy related constraints on these collaborations for legal, ethical, and competit...

متن کامل

PrivLogit: Efficient Privacy-preserving Logistic Regression by Tailoring Numerical Optimizers

Safeguarding privacy in machine learning is highly desirable, especially in collaborative studies across many organizations. Privacy-preserving distributed machine learning (based on cryptography) is popular to solve the problem. However, existing cryptographic protocols still incur excess computational overhead. Here, we make a novel observation that this is partially due to naive adoption of ...

متن کامل

The Large Margin Mechanism for Differentially Private Maximization

A basic problem in the design of privacy-preserving algorithms is the private maximization problem: the goal is to pick an item from a universe that (approximately) maximizes a data-dependent function, all under the constraint of differential privacy. This problem has been used as a sub-routine in many privacy-preserving algorithms for statistics and machine-learning. Previous algorithms for th...

متن کامل

Privacy Preserving Multi-party Machine Learning with Homomorphic Encryption

Privacy preserving multi-party machine learning approaches enable multiple parties to train a machine learning model from aggregate data while ensuring the privacy of their individual datasets is preserved. In this paper, we propose a privacy preserving multi-party machine learning approach based on homomorphic encryption where the machine learning algorithm of choice is deep neural networks. W...

متن کامل

Differentially Private Empirical Risk Minimization

Privacy-preserving machine learning algorithms are crucial for the increasingly common setting in which personal data, such as medical or financial records, are analyzed. We provide general techniques to produce privacy-preserving approximations of classifiers learned via (regularized) empirical risk minimization (ERM). These algorithms are private under the ε-differential privacy definition du...

متن کامل

Statement of Research — Alexandre Evfimievski

My prior research has been mainly in the area of privacy preserving data mining. It included such topics as: using randomization for preserving privacy of individual transactions in association rule mining; secure two-party computation of joins between two relational tables, set intersections, join sizes, and supports of vertically partitioned itemsets; improving space and time efficiency in pr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2015